# Long sequence processing
Codesearch ModernBERT Owl 2.0 Plus
Apache-2.0
The latest pre-trained model designed for high-quality code understanding and semantic retrieval, supporting long sequence processing of 8 programming languages.
Text Embedding
C
Shuu12121
602
1
Devstral Small 2505.w4a16 Gptq
Apache-2.0
This is a 4-bit GPTQ quantized version based on the mistralai/Devstral-Small-2505 model, optimized for consumer-grade hardware.
Large Language Model
Safetensors
D
mratsim
557
1
Codesearch ModernBERT Owl
Apache-2.0
A sentence transformer model specifically designed for code search, based on ModernBERT architecture, supporting sequence lengths of up to 2048 tokens
Text Embedding English
C
Shuu12121
75
1
Isoformer
Isoformer is a multimodal biological sequence model capable of processing DNA, RNA, and protein sequences, and predicting gene expression.
Protein Model
Transformers

I
isoformer-anonymous
21
2
Phi 2 Super GGUF
MIT
phi-2-super-GGUF is the GGUF quantized version of the abacaj/phi-2-super model, suitable for local execution and text generation tasks.
Large Language Model
Transformers

P
MaziyarPanahi
158
5
Mamba 370m Hf
Mamba is an efficient language model based on the State Space Model (SSM), with the ability to model sequences with linear time complexity.
Large Language Model
Transformers

M
state-spaces
6,895
14
Sauerkrautlm Mixtral 8x7B GGUF
Apache-2.0
SauerkrautLM Mixtral 8X7B is a multilingual text generation model based on the Mixtral architecture. It has been fine-tuned and aligned using SFT and DPO, and supports English, German, French, Italian, and Spanish.
Large Language Model
Transformers Supports Multiple Languages

S
TheBloke
403
8
Mythalion Kimiko V2 AWQ
Other
Mythalion Kimiko v2 - AWQ is the AWQ quantized version of the Mythalion Kimiko v2 model created by nRuaif, featuring high efficiency, accuracy, and fast inference.
Large Language Model
Transformers

M
TheBloke
403
7
Mistral 7B Instruct V0.1 GPTQ
Apache-2.0
The GPTQ quantized version of Mistral 7B Instruct v0.1, supporting operation under the ExLlama or Transformers framework
Large Language Model
Transformers

M
TheBloke
7,481
82
Codefuse 13B
Other
CodeFuse-13B is a 13-billion-parameter code generation model trained on the GPT-NeoX framework, supporting over 40 programming languages and capable of processing code sequences up to 4096 characters long.
Large Language Model
Transformers

C
codefuse-ai
31
49
Kogpt J 350m
MIT
A Korean text generation model based on the GPT-J architecture with 350 million parameters, suitable for various Korean text generation tasks.
Large Language Model Korean
K
heegyu
123
7
Long T5 Tglobal Xl
Apache-2.0
LongT5 is a Transformer-based text-to-text model specifically designed to handle long sequence inputs, supporting up to 16,384 tokens.
Large Language Model
Transformers English

L
google
336
23
Long T5 Local Large
Apache-2.0
Long T5 is a text-to-text Transformer model extended from T5, which supports efficient processing of long sequence inputs and is particularly suitable for text generation tasks.
Large Language Model
Transformers English

L
google
177
5
Xlarge
Apache-2.0
Funnel Transformer is an English text pre-training model based on self-supervised learning, adopting objectives similar to ELECTRA, achieving efficient language processing by filtering sequence redundancy.
Large Language Model
Transformers English

X
funnel-transformer
31
1
Bertin Roberta Large Spanish
BERTIN is a series of Spanish language models based on BERT. This model follows the RoBERTa-large architecture, trained from scratch using the Flax framework, with data sourced from the Spanish portion of the mC4 corpus.
Large Language Model Spanish
B
flax-community
26
0
Reformer Enwik8
A character-level Reformer language model trained on the enwik8 dataset, designed for text generation and compression tasks.
Large Language Model
Transformers

R
google
637
19
Featured Recommended AI Models